Comparison results for segmental training algorithms for mixture density HMMs

نویسنده

  • Mikko Kurimo
چکیده

This work presents experiments on four segmental training algorithms for mixture density HMMs. The segmental versions of SOM and LVQ3 suggested by the author are compared against the conventional segmental K-means and the segmental GPD. The recognition task used as a test bench is the speaker dependent, but vocabulary independent automatic speech recognition. The output density function of each state in each model is a mixture of multivariate Gaussian densities. Neural network methods SOM and LVQ are applied to learn the parameters of the density models from the mel-cepstrum features of the training samples. The segmental training improves the segmentation and the model parameters by turns to obtain the best possible result, because the segmentation and the segment classification depend on each other. It suffices to start the training process by dividing the training samples approximatively into phoneme samples.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Segmental Lvq Training for Phoneme Wise Tied Mixture Density Hmms

This work presents training methods and recogni tion experiments for phoneme wise tied mixture den sities in hidden Markov models HMM The system trains speaker dependent but vocabulary independent phoneme models for the recognition of Finnish words The Learning Vector Quantization LVQ methods are applied to increase the discrimination between the phoneme models A segmental LVQ training is pro p...

متن کامل

Self-organization in mixture densities of HMM based speech recognition

In this paper experiments are presented to apply Self-Organizing Map (SOM) and Learning Vector Quantization (LVQ) for training mixture density hidden Markov models (HMMs) in automatic speech recognition. The decoding of spoken words into text is made using speaker dependent, but vocabulary and context independent phoneme HMMs. Each HMM has a set of states and the output density of each state is...

متن کامل

Using the self-organizing map to speed up the probability density estimation for speech recognition with mixture density HMMs

This paper presents methods to improve the probability density estimation in hidden Markov models for phoneme recognition by exploiting the Self-Organizing Map (SOM) algorithm. The advantage of using the SOM is based on the created approximative topology between the mixture densities by training the Gaussian mean vectors used as the kernel centers by the SOM algorithm. The topology makes the ne...

متن کامل

Reviving discrete HMMs: the myth about the superiority of continuous HMMs

Despite what is generally believed, we have recently shown that discrete-distribution HMMs can outperform continuousdensity HMMs at significantly faster decoding speeds. Recognition performance and decoding speed of the discrete HMMs are improved by using product-code Vector Quantization (VQ) and mixtures of discrete distributions. In this paper, we present efficient training and decoding algor...

متن کامل

Efficient speech recognition using subvector quantization and discrete-mixture HMMs

This paper introduces a new form of observation distributions for hidden Markov models (HMMs), combining subvector quantization and mixtures of discrete distributions. We present efficient training and decoding algorithms for the discretemixture HMMs (DMHMMs). Our experimental results in the airtravel information domain show that the high-level of recognition accuracy of continuous mixture-dens...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1997